Identification of High-level Object Manipulation Operations from Multimodal Input
نویسنده
چکیده
Object manipulation constitutes a large part of our daily hand movements. Recognition of such movements by a robot in an interactive scenario is an issue that is rapidly gaining attention. In this paper we present an approach to identification of a class of high-level manual object manipulations. Experiments have shown that the naive approach based on classification of low-level sensor data yields poor performance. In this paper we introduce a two-stage procedure that considerably improves the identification performance. In the first stage of the procedure we estimate an intermediate representation by applying a linear preprocessor to the multimodal low-level sensor data. This mapping calculates shape, orientation and weight estimators of the interaction object. In the second stage we generate a classifier that is trained to identify high-level object manipulations given the intermediate representation based on shape, orientation and weight. The devices used in our procedure are: Immersion CyberGlove II enhanced with five tactile sensors on the fingertips (TouchGlove), nine tactile sensors to measure the change of the object’s weight and a VICON multicamera system for trajectory recording. We have achieved the following recognition rates for 3600 data samples representing a sequence of manual object manipulations: 100% correct labelling of “holding”, 97% of “pouring”, 81% of “squeezing” and 65% of “tilting”.
منابع مشابه
Unsupervised Segmentation of Object Manipulation Operations from Multimodal Input
We propose a procedure for unsupervised identification of bimanual high-level object manipulation operations in multimodal data. The presented procedure applies a two-stage segmentation and a selection step to observation sequences. We employ an unsupervised Bayesian segmentation method to identify homogeneous segments which correspond to primitive object manipulation operations. The data is re...
متن کاملTangible Objects for the Acquisition of Multimodal Interaction Patterns
Multimodal user interfaces offer more intuitive interaction for end-users, however, usually only through predefined input schemes. This paper describes a user experiment for multimodal interaction pattern identification, using head gesture and speech inputs for a 3D graph manipulation. We show that a direct mapping between head gestures and the 3D object predominates, however even for such a si...
متن کاملPlanning and Control of Two-Link Rigid Flexible Manipulators in Dynamic Object Manipulation Missions
This research focuses on proposing an optimal trajectory planning and control method of two link rigid-flexible manipulators (TLRFM) for Dynamic Object Manipulation (DOM) missions. For the first time, achievement of DOM task using a rotating one flexible link robot was taken into account in [20]. The authors do not aim to contribute on either trajectory tracking or vibration control of the End-...
متن کاملManipulation As an Ideological Tool in the Persian Translations of Ervand Abrahamian’s The Coup: A Multimodal CDA Approach
The present Critical Discourse Analysis (CDA) study aimed to explore the probable ideological manipu- lations exerted in three translations of an English political book entitled The Coup by Ervand Abraha- mian. This comparative qualitative study was conducted based on Farahzad‘s three-dimensional CDA model. The textual, paratextual, and ...
متن کاملA Communicative Mediator in a Virtual Environment: Processing of Multimodal Input and Output
1 Abstract This paper presents work on multimodal communication with an anthropomorphic agent. It focuses on processing of multimodal input and output employing natural language and gestures in virtual environments. On the input side, we describe our approach to recognize and interpret co-verbal gestures used for pointing, object manipulation, and object description. On the output side, we pres...
متن کامل